Contextual Inverse Optimization: Offline and Online Learning
نویسندگان
چکیده
We study the problems of offline and online contextual optimization with feedback information, where instead observing loss, we observe, after-the-fact, optimal action an oracle full knowledge objective function would have taken. aim to minimize regret, which is defined as difference between our losses ones incurred by all-knowing oracle. In setting, decision-maker has information available from past periods needs make one decision, while in optimizes decisions dynamically over time based a new set feasible actions functions each period. For characterize minimax policy, establishing performance that can be achieved underlying geometry induced data. leverage this geometric characterization optimize cumulative regret. develop algorithm yields first regret bound for problem logarithmic horizon.
منابع مشابه
Offline Optimization for Online Ad Allocation
We consider online ad allocation from the perspective of optimizing delivery of a given set of ad reservations. Ad allocation naturally fits into the class of online bipartite matching problems: ad nodes are fixed, and impression (ad slot) nodes arrive one at a time, and must be assigned a qualified ad upon arrival. However most previous work on online matching does not model an important aspec...
متن کاملOnline Optimization through Mining the Offline Optimum
Ports, warehouses and courier services have to decide online how an arriving task is to be served in order that cost is minimized (or profit maximized). These operators have a wealth of historical data on task assignments; can these data be mined for knowledge or rules that can help the decisionmaking? MOO is a novel application of data mining to online optimization. The idea is to mine (logged...
متن کاملEmulating the Expert: Inverse Optimization through Online Learning
In this paper, we demonstrate how to learn the objective function of a decision maker while only observing the problem input data and the decision maker’s corresponding decisions over multiple rounds. Our approach is based on online learning techniques and works for linear objectives over arbitrary sets for which we have a linear optimization oracle and as such generalizes previous work based o...
متن کاملOnline Heterogeneous Transfer Learning by Weighted Offline and Online Classifiers
In this paper, we study online heterogeneous transfer learning (HTL) problems where offline labeled data from a source domain is transferred to enhance the online classification performance in a target domain. The main idea of our proposed algorithm is to build an offline classifier based on heterogeneous similarity constructed by using labeled data from a source domain and unlabeled co-occurre...
متن کاملOffline Evaluation of Online Reinforcement Learning Algorithms
In many real-world reinforcement learning problems, we have access to an existing dataset and would like to use it to evaluate various learning approaches. Typically, one would prefer not to deploy a fixed policy, but rather an algorithm that learns to improve its behavior as it gains more experience. Therefore, we seek to evaluate how a proposed algorithm learns in our environment, meaning we ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Social Science Research Network
سال: 2021
ISSN: ['1556-5068']
DOI: https://doi.org/10.2139/ssrn.3863366